18 research outputs found

    Listener expectations and the perceptual accommodation of talker variability: A pre-registered replication

    Get PDF
    Published: 04 May 2021Researchers have hypothesized that in order to accommodate variability in how talkers produce their speech sounds, listeners must perform a process of talker normalization. Consistent with this proposal, several studies have shown that spoken word recognition is slowed when speech is produced by multiple talkers compared with when all speech is produced by one talker (a multitalker processing cost). Nusbaum and colleagues have argued that talker normalization is modulated by attention (e.g., Nusbaum & Morin, 1992, Speech Perception, Production and Linguistic Structure, pp. 113–134). Some of the strongest evidence for this claim is from a speeded monitoring study where a group of participants who expected to hear two talkers showed a multitalker processing cost, but a separate group who expected one talker did not (Magnuson & Nusbaum, 2007, Journal of Experimental Psychology, 33[2], 391–409). In that study, however, the sample size was small and the crucial interaction was not significant. In this registered report, we present the results of a well-powered attempt to replicate those findings. In contrast to the previous study, we did not observe multitalker processing costs in either of our groups. To rule out the possibility that the null result was due to task constraints, we conducted a second experiment using a speeded classification task. As in Experiment 1, we found no influence of expectations on talker normalization, with no multitalker processing cost observed in either group. Our data suggest that the previous findings of Magnuson and Nusbaum (2007) be regarded with skepticism and that talker normalization may not be permeable to high-level expectations.This research was supported by NSF 1754284, NSF IGERT 1144399 & NSF NRT 1747486 (PI: JSM) and NSF BCS 1554810 & NIH R01 DC013064 (PI: EBM). This research was also supported in part by the Basque Government through the BERC 2018- 2021 program and by the Agencia Estatal de Investigación through BCBL Severo Ochoa excellence accreditation SEV-2015-0490. SL was supported by an NSF Graduate Research Fellowshi

    Neural substrates of subphonemic variation and lexical competition in spoken word recognition

    Get PDF
    In spoken word recognition, subphonemic variation influences lexical activation, with sounds near a category boundary increasing phonetic competition as well as lexical competition. The current study investigated the interplay of these factors using a visual world task in which participants were instructed to look at a picture of an auditory target (e.g. peacock). Eyetracking data indicated that participants were slowed when a voiced onset competitor (e.g. beaker) was also displayed, and this effect was amplified when acoustic-phonetic competition was increased. Simultaneously-collected fMRI data showed that several brain regions were sensitive to the presence of the onset competitor, including the supramarginal, middle temporal, and inferior frontal gyri, and functional connectivity analyses revealed that the coordinated activity of left frontal regions depends on both acoustic-phonetic and lexical factors. Taken together, results suggest a role for frontal brain structures in resolving lexical competition, particularly as atypical acoustic-phonetic information maps on to the lexicon.Research was supported by National Institutes of Health (NIH) [grant number: R01 DC013064] to EBM and NIH NIDCD [grant number R01 DC006220] to SEB. SG was supported by the Spanish Ministry of Economy and Competitiveness through the Severo Ochoa Programme for Centres/Units of Excellence in R&D [SEV‐2015‐490]. The contents of this paper reflect the views of the authors and not those of the funding agencies

    Does signal reduction imply predictive coding in models of spoken word recognition?

    Get PDF
    Published online: 14 April 2021Pervasive behavioral and neural evidence for predictive processing has led to claims that language processing depends upon predictive coding. Formally, predictive coding is a computational mechanism where only deviations from top-down expectations are passed between levels of representation. In many cognitive neuroscience studies, a reduction of signal for expected inputs is taken as being diagnostic of predictive coding. In the present work, we show that despite not explicitly implementing prediction, the TRACE model of speech perception exhibits this putative hallmark of predictive coding, with reductions in total lexical activation, total lexical feedback, and total phoneme activation when the input conforms to expectations. These findings may indicate that interactive activation is functionally equivalent or approximant to predictive coding or that caution is warranted in interpreting neural signal reduction as diagnostic of predictive coding.This researchwas supported by NSF 1754284, NSF IGERT 1144399, and NSF NRT 1747486 (PI: J.S.M.). This research was also supported in part by the Basque Government through the BERC 2018- 2021program, and by the Agencia Estatal de Investigación through BCBL Severo Ochoa excellenceaccreditation SEV-2015-0490. S.L. was supported by an NSF Graduate Research Fellowship

    Lexical Information Guides Retuning of Neural Patterns in Perceptual Learning for Speech

    Get PDF
    Posted Online August 31, 2020A listener's interpretation of a given speech sound can vary probabilistically from moment to moment. Previous experience (i.e., the contexts in which one has encountered an ambiguous sound) can further influence the interpretation of speech, a phenomenon known as perceptual learning for speech. This study used multivoxel pattern analysis to query how neural patterns reflect perceptual learning, leveraging archival fMRI data from a lexically guided perceptual learning study conducted by Myers and Mesite [Myers, E. B., & Mesite, L. M. Neural systems underlying perceptual adjustment to non-standard speech tokens. Journal of Memory and Language, 76, 80-93, 2014]. In that study, participants first heard ambiguous /s/-/∫/ blends in either /s/-biased lexical contexts (epi_ode) or /∫/-biased contexts (refre_ing); subsequently, they performed a phonetic categorization task on tokens from an /asi/-/a∫i/ continuum. In the current work, a classifier was trained to distinguish between phonetic categorization trials in which participants heard unambiguous productions of /s/ and those in which they heard unambiguous productions of /∫/. The classifier was able to generalize this training to ambiguous tokens from the middle of the continuum on the basis of individual participants' trial-by-trial perception. We take these findings as evidence that perceptual learning for speech involves neural recalibration, such that the pattern of activation approximates the perceived category. Exploratory analyses showed that left parietal regions (supramarginal and angular gyri) and right temporal regions (superior, middle, and transverse temporal gyri) were most informative for categorization. Overall, our results inform an understanding of how moment-to-moment variability in speech perception is encoded in the brain.This work was supported by NSF IGERT DGE-1144399, NIH R03 DC009395 (PI: Myers), NIH R01 DC013064 (PI: Myers), and an NSF Graduate Research Fellowship to S. L. The authors report no conflict of interes

    Using TMS to evaluate a causal role for right posterior temporal cortex in talker-specific phonetic processing

    Get PDF
    Available online 21 April 2023Theories suggest that speech perception is informed by listeners’ beliefs of what phonetic variation is typical of a talker. A previous fMRI study found right middle temporal gyrus (RMTG) sensitivity to whether a phonetic variant was typical of a talker, consistent with literature suggesting that the right hemisphere may play a key role in conditioning phonetic identity on talker information. The current work used transcranial magnetic stimulation (TMS) to test whether the RMTG plays a causal role in processing talker-specific phonetic variation. Listeners were exposed to talkers who differed in how they produced voiceless stop consonants while TMS was applied to RMTG, left MTG, or scalp vertex. Listeners subsequently showed near-ceiling performance in indicating which of two variants was typical of a trained talker, regardless of previous stimulation site. Thus, even though the RMTG is recruited for talker-specific phonetic processing, modulation of its function may have only modest consequences.This research was supported by NSF 1554810 (PI: EBM), NIH NIDCD 2R01 DC013064 (PI: EBM) and NSF NRT 1747486 (PI: JSM). This research was supported in part by the Basque Government through the BERC 2022–2025 program, by the Spanish State Research Agency through BCBL Severo Ochoa excellence accreditation CEX2020-001010- S and award PID2020-119131 GB-I000

    Robust Lexically Mediated Compensation for Coarticulation: Christmash Time Is Here Again

    Get PDF
    First published: 20 April 2021A long-standing question in cognitive science is how high-level knowledge is integrated with sensory input. For example, listeners can leverage lexical knowledge to interpret an ambiguous speech sound, but do such effects reflect direct top-down influences on perception or merely postperceptual biases? A critical test case in the domain of spoken word recognition is lexically mediated compensation for coarticulation (LCfC). Previous LCfC studies have shown that a lexically restored context phoneme (e.g., /s/ in Christma#) can alter the perceived place of articulation of a subsequent target phoneme (e.g., the initial phoneme of a stimulus from a tapes-capes continuum), consistent with the influence of an unambiguous context phoneme in the same position. Because this phoneme-to-phoneme compensation for coarticulation is considered sublexical, scientists agree that evidence for LCfC would constitute strong support for top–down interaction. However, results from previous LCfC studies have been inconsistent, and positive effects have often been small. Here, we conducted extensive piloting of stimuli prior to testing for LCfC. Specifically, we ensured that context items elicited robust phoneme restoration (e.g., that the final phoneme of Christma# was reliably identified as /s/) and that unambiguous context-final segments (e.g., a clear /s/ at the end of Christmas) drove reliable compensation for coarticulation for a subsequent target phoneme.We observed robust LCfC in a well-powered, preregistered experiment with these pretested items (N = 40) as well as in a direct replication study (N = 40). These results provide strong evidence in favor of computational models of spoken word recognition that include top–down feedback

    Brain-behavior relationships in incidental learning of non-native phonetic categories

    Get PDF
    Available online 12 September 2019.Research has implicated the left inferior frontal gyrus (LIFG) in mapping acoustic-phonetic input to sound category representations, both in native speech perception and non-native phonetic category learning. At issue is whether this sensitivity reflects access to phonetic category information per se or to explicit category labels, the latter often being required by experimental procedures. The current study employed an incidental learning paradigm designed to increase sensitivity to a difficult non-native phonetic contrast without inducing explicit awareness of the categorical nature of the stimuli. Functional MRI scans revealed frontal sensitivity to phonetic category structure both before and after learning. Additionally, individuals who succeeded most on the learning task showed the largest increases in frontal recruitment after learning. Overall, results suggest that processing novel phonetic category information entails a reliance on frontal brain regions, even in the absence of explicit category labels.This research was supported by NIH grant R01 DC013064 to EBM and NIH NIDCD Grant R01 DC006220 to SEB. The authors thank F. Sayako Earle for assistance with stimulus development; members of the Language and Brain lab for help with data collection and their feedback throughout the project; Elisa Medeiros for assistance with collection of fMRI data; Paul Taylor for assistance with neuroimaging analyses; and attendees of the 2016 Meeting of the Psychonomic Society and the 2017 Meeting of the Society for Neurobiology of Language for helpful feedback on this project. We also extend thanks to two anonymous reviewers for helpful feedback on a previous version of this manuscript

    The Influence of Sentence Context on Phonetic Recalibration

    Get PDF
    Individual talkers vary considerably in how they produce different speech sounds, and a challenge for the listener is to learn the appropriate mapping between acoustics and phonetic categories for an individual talker. Several studies have shown that listeners are able to leverage various sources of context (e.g., coincident visual information, lexical knowledge) to guide this process, sometimes termed perceptual learning of speech. Here, we examine how sentence-level semantic information – specifically, whether preceding sentence context is predictive of an upcoming word – might modulate the size of learning effects. Across a series of perceptual learning experiments, we manipulate how learning compares between groups who receive neutral or predictive sentence contexts, also varying whether contexts are presented in the auditory or written modality. Though we observed greater learning for subjects who read predictive contexts than for subjects who read neutral contexts, this finding did not replicate in an identical follow-up experiment, suggesting that potential influences of sentence context on phonetic recalibration may be small. These findings are discussed in the context of the broader literature on perceptual learning

    TimedUCM: Specifying scenario models over time

    No full text
    This thesis introduces TimedUCM (Timed Use Case Maps), an extension of the current UCM (Use Case Maps) standard, which enables the modeling and analysis of a comprehensive set of changes to a scenario model over time. Scenario models are used to capture functional requirements in terms of causal relationships among scenario steps and combine scenarios into a high-level system view where behavior is superimposed over structural elements. Scenarios are analyzed with a traversal mechanism that checks whether expected pre- and post-conditions of a scenario hold when the model is updated. Generally, these analyses focus on one snapshot in time, i.e., the model is assumed to not change over time. However, desired system qualities, stakeholder objectives, potential solutions, and their relationships as well as other model elements may change as the time progresses, and sometimes this evolution over time is predictable or can be reasonably estimated. It may also be useful to explore several "what if" scenarios in terms of how the system model could evolve. Furthermore, some systems may require the problem to be analyzed over a longer period of time.TimedUCM proposes capturing the scenario model and the changes to the scenario model in a single model to facilitate system evolution. The metamodel for TimedURN (Timed User Requirements Notation) is presented and the usefulness of TimedUCM is illustrated with a hypothetical but realistic example to build an advanced research facility, and the comprehensiveness of the supported types of changes is implemented and assessed in jUCMNav, a requirements engineering tool. Moreover, a detailed case study has been performed to show TimedUCM's functionality and usefulness.Cette thèse présente TimedUCM (Timed Use Case Maps), une extension de la norme actuelle UCM (Use Case Maps), qui permet de modéliser et d'analyser un ensemble complet de modifications d'un modèle de scénario dans le temps. Les modèles de scénarios sont utilisés pour capturer les exigences fonctionnelles en termes de relations causales entre les étapes d'un scénario, et combiner les scénarios dans une vue système de haut niveau où le comportement est superposé aux éléments structurels. Les scénarios sont analysés avec un mécanisme de traversée qui vérifie si les pré- et post-conditions d'un scénario sont maintenues lorsque le modèle est mis à jour. Généralement, ces analyses se concentrent sur un moment particulier dans le temps, c'est-à-dire que le modèle est supposé ne pas changer avec le temps. Cependant, les qualités souhaitées du système, les objectifs des parties prenantes, les solutions potentielles et leurs relations ainsi que d'autres éléments du modèle peuvent changer avec le temps, et parfois cette évolution est prévisible ou peut être raisonnablement estimée. Il peut également être utile d'explorer plusieurs scénarios anticipatifs (what if) en termes d'évolution du modèle du système. De plus, certains systèmes peuvent exiger que le problème soit analysé sur une période plus longue.TimedUCM propose de capturer le modèle de scénario et les modifications apportées au modèle de scénario dans un seul modèle pour faciliter l'évolution du système. Le métamodèle pour TimedURN (Timed User Requirements Notation) est présenté et l'utilité de TimedUCM est illustrée avec un exemple hypothétique mais réaliste de construction d'une installation de recherche avancée, et l'exhaustivité des types de changements pris en charge est implémentée et évaluée dans jUCMNav, un outil d'ingénierie des exigences. De plus, une étude de cas détaillée a été réalisée pour montrer la fonctionnalité et l'utilité de TimedUCM
    corecore